Random walk

A random walk, sometimes denoted RW, is a mathematical formalisation of a trajectory that consists of taking successive random steps. For example, the path traced by a molecule as it travels in a liquid or a gas, the search path of a foraging animal, the price of a fluctuating stock and the financial status of a gambler can all be modeled as random walks. The term random walk was first introduced by Karl Pearson in 1905.[1] Random walks have been used in many fields: ecology, economics, psychology, computer science, physics, chemistry, and biology.[2][3][4][5][6][7][8][9] Random walks explain the observed behaviors of processes in these fields, and thus serve as a fundamental model for the recorded stochastic activity.

Various different types of random walks are of interest. Often, random walks are assumed to be Markov chains or Markov processes, but other, more complicated walks are also of interest. Some random walks are on graphs, others on the line, in the plane, or in higher dimensions, while some random walks are on groups. Random walks also vary with regard to the time parameter. Often, the walk is in discrete time, and indexed by the natural numbers, as in X_0,X_1,X_2,\dots. However, some walks take their steps at random times, and in that case the position X_t is defined for the continuum of times t\ge 0. Specific cases or limits of random walks include the drunkard's walk and Lévy flight. Random walks are related to the diffusion models and are a fundamental topic in discussions of Markov processes. Several properties of random walks, including dispersal distributions, first-passage times and encounter rates, have been extensively studied.

Contents

Lattice random walk

A popular random walk model is that of a random walk on a regular lattice, where at each step the walk jumps to another site according to some probability distribution. In simple random walk, the walk can only jump to neighbouring sites of the lattice. In simple symmetric random walk on a locally finite lattice, the probabilities of walk jumping to any one of its neighbours are the same. The most well-studied example is of random walk on the d-dimensional integer lattice (sometimes called the hypercubic lattice) \mathbb Z^d.

One-dimensional random walk

An elementary example of a random walk is the random walk on the integer number line, \mathbb Z, which starts at 0 and at each step moves +1 or −1 with equal probability.

This walk can be illustrated as follows. A marker is placed at zero on the number line and a fair coin is flipped. If it lands on heads, the marker is moved one unit to the right. If it lands on tails, the marker is moved one unit to the left. After five flips, it is possible to have landed on 1, −1, 3, −3, 5, or −5. With five flips, three heads and two tails, in any order, will land on 1. There are 10 ways of landing on 1 or −1 (by flipping three tails and two heads), 5 ways of landing on 3 (by flipping four heads and one tail), 5 ways of landing on −3 (by flipping four tails and one head), 1 way of landing on 5 (by flipping five heads), and 1 way of landing on −5 (by flipping five tails). See the figure below for an illustration of the possible outcomes of 5 flips.

To define this walk formally, take independent random variables Z_1, Z_2,\dots, where each variable is either 1 or −1, with a 50% probability for either value, and set S_0 = 0\,\! and S_n =\sum_{j=1}^nZ_j. The series \{S_n\}\,\! is called the simple random walk on \mathbb Z. This series (the sum of the sequence of −1s and 1s) gives the distance walked, if each part of the walk is of length one. The expectation E(S_n)\,\! of S_n\,\! is zero. That is, the mean of all coin flips approaches zero as the number of flips increase. This follows by the finite additivity property of expectation:

E(S_n)=\sum_{j=1}^n E(Z_j)=0.

A similar calculation, using the independence of the random variables and the fact that E(Z_n^2)=1, shows that:

E(S_n^2)=\sum_{j=1}^n E(Z_j^2)=n.

This hints that E(|S_n|)\,\!, the expected translation distance after n steps, should be of the order of \sqrt n. In fact,

\lim_{n\to\infty} \frac{E(|S_n|)}{\sqrt n}= \sqrt{\frac 2{\pi}}.

Derivation of dispersion proportionality

If we have the situation where the probabilities of moving either left or right are equal, p = 1/2 and q=1/2, the probability of taking k steps to the right out of a total of N steps is given by

P(k;N) = \binom{N}{k}\left(\frac{1}{2}\right)^N

since there are \binom{N}{k} possible ways of taking k and N-k steps to the right and left, respectively. The probability of taking any of these independent steps is 1/2, and so we have the product 1/2^N. Now, the expectation value of taking k steps is


\begin{align}
  <k> &= \sum_{k=0}^N k P(k;N) \\
   &= \sum_{k=0}^N k \binom{N}{k}\left(\frac{1}{2}\right)^N \\
   &= \left(\frac{1}{2}\right)^N \sum_{k=0}^N N \binom{N-1}{k-1} \\
   &= \left(\frac{1}{2}\right)^N N 2^{N-1} \\
   &= \frac{N}{2}.
\end{align}

It is generally the case that <k> = Np . Note the identity we have used with the binomial coefficient k\binom{N}{k}. We use it again below. We must then calculate the expectation value of k^2:


\begin{align}
  <k^2> &= \sum_{k=0}^N k^2 P(k;N) \\
   &= \left(\frac{1}{2}\right)^N \sum_{k=0}^N (k-1 %2B 1) N \binom{N-1}{k-1} \\
   &= \left(\frac{1}{2}\right)^N N\sum_{k=0}^N \left[ (N-1)\binom{N-2}{k-2} %2B \binom{N-1}{k-1} \right] \\
   &= \left(\frac{1}{2}\right)^N N \left[ (N-1)2^{N-2} %2B 2^{N-1} \right] \\
   &= \frac{N^2}{4} %2B \frac{N}{4}.
\end{align}

It is generally the case that <k^2> = (Np)^2 %2B Npq. The dispersion is


\begin{align}
  <\Delta k>^2 &= \sqrt{\langle k^2\rangle-\langle k\rangle^2} \\
   &\propto \sqrt{N}.
\end{align}

This result shows that diffusion becomes quite an ineffective way to mix solutions because of the way the square root behaves for large N.

How many times will a random walk cross a boundary line if permitted to continue walking forever? A simple random walk on \mathbb Z will cross every point an infinite number of times. This result has many names: the level-crossing phenomenon, recurrence or the gambler's ruin. The reason for the last name is as follows: a gambler with a finite amount of money will always lose when playing a fair game against a bank with an infinite amount of money. The gambler's money will perform a random walk, and it will reach zero at some point, and the game will be over.

If a and b are positive integers, then the expected number of steps until a one-dimensional simple random walk starting at 0 first hits b or −a is ab. The probability that this walk will hit b before −a is a/(a%2Bb), which can be derived from the fact that simple random walk is a martingale.

Some of the results mentioned above can be derived from properties of Pascal's triangle. The number of different walks of n steps where each step is +1 or −1 is clearly 2n. For the simple random walk, each of these walks are equally likely. In order for Sn to be equal to a number k it is necessary and sufficient that the number of +1 in the walk exceeds those of −1 by k. Thus, the number of walks which satisfy S_n=k is precisely the number of ways of choosing (n + k)/2 elements from an n element set (for this to be non-zero, it is necessary that n + k be an even number), which is an entry in Pascal's triangle denoted by n \choose (n%2Bk)/2. Therefore, the probability that S_n=k is equal to 2^{-n}{n\choose (n%2Bk)/2}. By representing entries of Pascal's triangle in terms of factorials and using Stirling's formula, one can obtain good estimates for these probabilities for large values of n.

This relation with Pascal's triangle is demonstrated for small values of n. At zero turns, the only possibility will be to remain at zero. However, at one turn, there is one chance of landing on −1 or one chance of landing on 1. At two turns, a marker at 1 could move to 2 or back to zero. A marker at −1, could move to −2 or back to zero. Therefore, there is one chance of landing on −2, two chances of landing on zero, and one chance of landing on 2.

n −5 −4 −3 −2 −1 0 1 2 3 4 5
P[S_0=k] 1
2P[S_1=k] 1 1
2^2P[S_2=k] 1 2 1
2^3P[S_3=k] 1 3 3 1
2^4P[S_4=k] 1 4 6 4 1
2^5P[S_5=k] 1 5 10 10 5 1

The central limit theorem and the law of the iterated logarithm describe important aspects of the behavior of simple random walk on \mathbb Z.

As a Markov chain

A one-dimensional random walk can also be looked at as a Markov chain whose state space is given by the integers i=0,\pm 1,\pm 2,\dots . For some number p satisying \,0 < p < 1, the transition probabilities (the probability Pi,j of moving from state i to state j) are given by

\,P_{i,i%2B1}=p=1-P_{i,i-1}.

Gaussian random walk

A random walk having a step size that varies according to a normal distribution is used as a model for real-world time series data such as financial markets. The Black–Scholes formula for modeling option prices, for example, uses a gaussian random walk as an underlying assumption.

Here, the step size is the inverse cumulative normal distribution \Phi^{-1}(z,\mu,\sigma) where 0 ≤ z ≤ 1 is a uniformly distributed random number, and μ and σ are the mean and standard deviations of the normal distribution, respectively.

For steps distributed according to any distribution with zero mean and a finite variance (not necessarily just a normal distribution), the root mean square translation distance after n steps is

\sqrt{E|S_n^2|} = \sigma \sqrt{n}.

Higher dimensions

Imagine now a drunkard walking randomly in an idealized city. The city is effectively infinite and arranged in a square grid, and at every intersection, the drunkard chooses one of the four possible routes (including the one he came from) with equal probability. Formally, this is a random walk on the set of all points in the plane with integer coordinates. Will the drunkard ever get back to his home from the bar? It turns out that he will. This is the high dimensional equivalent of the level crossing problem discussed above. The probability of returning to the origin decreases as the number of dimensions increases. In three dimensions, the probability decreases to roughly 34%. A derivation, along with values of p(d) are discussed here: Pólya's Random Walk Constants.

The trajectory of a random walk is the collection of sites it visited, considered as a set with disregard to when the walk arrived at the point. In one dimension, the trajectory is simply all points between the minimum height the walk achieved and the maximum (both are, on average, on the order of √n). In higher dimensions the set has interesting geometric properties. In fact, one gets a discrete fractal, that is a set which exhibits stochastic self-similarity on large scales, but on small scales one can observe "jaggedness" resulting from the grid on which the walk is performed. The two books of Lawler referenced below are a good source on this topic.

Relation to Wiener process

A Wiener process is a stochastic process with similar behaviour to Brownian motion, the physical phenomenon of a minute particle diffusing in a fluid. (Sometimes the Wiener process is called "Brownian motion", although this is strictly speaking a confusion of a model with the phenomenon being modeled.)

A Wiener process is the scaling limit of random walk in dimension 1. This means that if you take a random walk with very small steps you get an approximation to a Wiener process (and, less accurately, to Brownian motion). To be more precise, if the step size is ε, one needs to take a walk of length L2 to approximate a Wiener process walk of length L. As the step size tends to 0 (and the number of steps increases proportionally) random walk converges to a Wiener process in an appropriate sense. Formally, if B is the space of all paths of length L with the maximum topology, and if M is the space of measure over B with the norm topology, then the convergence is in the space M. Similarly, a Wiener process in several dimensions is the scaling limit of random walk in the same number of dimensions.

A random walk is a discrete fractal (a function with integer dimensions; 1, 2, ...), but a Wiener process trajectory is a true fractal, and there is a connection between the two. For example, take a random walk until it hits a circle of radius r times the step length. The average number of steps it performs is r2. This fact is the discrete version of the fact that a Wiener process walk is a fractal of Hausdorff dimension 2 .[10] In two dimensions, the average number of points the same random walk has on the boundary of its trajectory is r4/3. This corresponds to the fact that the boundary of the trajectory of a Wiener process is a fractal of dimension 4/3, a fact predicted by Mandelbrot using simulations but proved only in 2000 by Lawler, Schramm and Werner [11].

A Wiener process enjoys many symmetries random walk does not. For example, a Wiener process walk is invariant to rotations, but random walk is not, since the underlying grid is not (random walk is invariant to rotations by 90 degrees, but Wiener processes are invariant to rotations by, for example, 17 degrees too). This means that in many cases, problems on random walk are easier to solve by translating them to a Wiener process, solving the problem there, and then translating back. On the other hand, some problems are easier to solve with random walks due to its discrete nature.

Random walk and Wiener process can be coupled, namely manifested on the same probability space in a dependent way that forces them to be quite close. The simplest such coupling is the Skorokhod embedding, but other, more precise couplings exist as well.

The convergence of a random walk toward the Wiener process is controlled by the central limit theorem. For a particle in a known fixed position at t = 0, the theorem tells us that after a large number of independent steps in the random walk, the walker's position is distributed according to a normal distribution of total variance:

\sigma^2 = \frac{t}{\delta t}\,\varepsilon^2,

where t is the time elapsed since the start of the random walk, \varepsilon is the size of a step of the random walk, and \delta t is the time elapsed between two successive steps.

This corresponds to the Green function of the diffusion equation that controls the Wiener process, which demonstrates that, after a large number of steps, the random walk converges toward a Wiener process.

In 3D, the variance corresponding to the Green's function of the diffusion equation is:

\sigma^2 = 6\,D\,t

By equalizing this quantity with the variance associated to the position of the random walker, one obtains the equivalent diffusion coefficient to be considered for the asymptotic Wiener process toward which the random walk converges after a large number of steps:

D = \frac{\varepsilon^2}{6 \delta t} (valid only in 3D)

Remark: the two expressions of the variance above correspond to the distribution associated to the vector \vec R that links the two ends of the random walk, in 3D. The variance associated to each component R_x, R_y or R_z is only one third of this value (still in 3D).

Anomalous diffusion

In disordered systems such as porous media and fractals \sigma^2 may not be proportional to t but to t^{2 / d_w}. The exponent d_w is called the anomalous diffusion exponent and can be larger or smaller than 2.[12]

Applications

The following are some applications of random walk:

In all these cases, random walk is often substituted for Brownian motion.

  • When this last approach is used in computer science it is known as Markov Chain Monte Carlo or MCMC for short. Often, sampling from some complicated state space also allows one to get a probabilistic estimate of the space's size. The estimate of the permanent of a large matrix of zeros and ones was the first major problem tackled using this approach.

Variants of random walks

A number of types of stochastic processes have been considered that are similar to the pure random walks but where the simple structure is allowed to be more generalized. The pure structure can be characterized by the steps being being defined by independent and identically distributed random variables.

Random walk on graphs

A random walk of length k on a possibly infinite graph G with a root 0 is a stochastic process with random variables X_1,X_2,\dots,X_k such that X_1=0 and  {X_{i%2B1}} is a vertex chosen uniformly at random from the neighbors of X_i. Then the number p_{v,w,k}(G) is the probability that a random walk of length k starting at v ends at w. In particular, if G is a graph with root 0, p_{0,0,2k} is the probability that a 2k-step random walk returns to 0.

Assume now that our city is no longer a perfect square grid. When our drunkard reaches a certain junction he picks between the various available roads with equal probability. Thus, if the junction has seven exits the drunkard will go to each one with probability one seventh. This is a random walk on a graph. Will our drunkard reach his home? It turns out that under rather mild conditions, the answer is still yes. For example, if the lengths of all the blocks are between a and b (where a and b are any two finite positive numbers), then the drunkard will, almost surely, reach his home. Notice that we do not assume that the graph is planar, i.e. the city may contain tunnels and bridges. One way to prove this result is using the connection to electrical networks. Take a map of the city and place a one ohm resistor on every block. Now measure the "resistance between a point and infinity". In other words, choose some number R and take all the points in the electrical network with distance bigger than R from our point and wire them together. This is now a finite electrical network and we may measure the resistance from our point to the wired points. Take R to infinity. The limit is called the resistance between a point and infinity. It turns out that the following is true (an elementary proof can be found in the book by Doyle and Snell):

Theorem: a graph is transient if and only if the resistance between a point and infinity is finite. It is not important which point is chosen if the graph is connected.

In other words, in a transient system, one only needs to overcome a finite resistance to get to infinity from any point. In a recurrent system, the resistance from any point to infinity is infinite.

This characterization of recurrence and transience is very useful, and specifically it allows us to analyze the case of a city drawn in the plane with the distances bounded.

A random walk on a graph is a very special case of a Markov chain. Unlike a general Markov chain, random walk on a graph enjoys a property called time symmetry or reversibility. Roughly speaking, this property, also called the principle of detailed balance, means that the probabilities to traverse a given path in one direction or in the other have a very simple connection between them (if the graph is regular, they are just equal). This property has important consequences.

Starting in the 1980s, much research has gone into connecting properties of the graph to random walks. In addition to the electrical network connection described above, there are important connections to isoperimetric inequalities, see more here, functional inequalities such as Sobolev and Poincaré inequalities and properties of solutions of Laplace's equation. A significant portion of this research was focused on Cayley graphs of finitely generated groups. For example, the proof of Dave Bayer and Persi Diaconis that 7 riffle shuffles are enough to mix a pack of cards (see more details under shuffle) is in effect a result about random walk on the group Sn, and the proof uses the group structure in an essential way. In many cases these discrete results carry over to, or are derived from manifolds and Lie groups.

A good reference for random walk on graphs is the online book by Aldous and Fill. For groups see the book of Woess. If the transition kernel p(x,y) is itself random (based on an environment \omega) then the random walk is called a "random walk in random environment". When the law of the random walk includes the randomness of \omega, the law is called the annealed law; on the other hand, if \omega is seen as fixed, the law is called a quenched law. See the book of Hughes or the lecture notes of Zeitouni.

We can think about choosing every possible edge with the same probability as maximizing uncertainty (entropy) locally. We could also do it globally – in maximal entropy random walk (MERW) we want all paths to be equally probable, or in other words: for each two vertexes, each path of given length is equally probable. This random walk has much stronger localization properties.

Self-interacting random walks

There are a number of interesting models of random paths in which each step depends on the past in a complicated manner. All are more complex for solving analytically than the usual random walk; still, the behavior of any model of a random walker is obtainable using computer. Examples include:

Correlated steps

The above descriptions assume that the steps in the random walk are independent. A variant of the model allows for correlated steps. This results in more steps being taken in either direction, and leads to larger displacements from the origin. Specifically, when the steps are independent, the expected displacement after n steps is \sqrt{n} = n^{0.5}. If the steps are long-range dependent (called also long-range correlated), however, the expected displacement becomes n^H for 0.5 < H < 1. H is the Hurst parameter (called also Hurst exponent).

Long range correlated time series is found in many biological, climatological and economic systems, including heartbeat, temperature records and volatility of stocks.

Long-range correlated walks

Long-range correlated time series are found in many biological, climatological and economic systems.

Heterogeneous random walks in one dimension

Heterogeneous random walks in one dimension can have either discrete time or continuous time. The interval is also either discrete or continuous, and it is either finite or without bounds. In a discrete system, the connections are among adjacent states. The dynamics are either Markovian, Semi-Markovian, or even not-Markovian depending on the model. Heterogeneous random walks in 1D have jump probabilities that depend on the location in the system, and/or different jumping time (JT) probability density functions (PDFs) that depend on the location in the system.
Known important results in simple systems include:

In a completely heterogeneous semi Markovian random walk in a discrete system of L (>1) states, the Green's function was found in Laplace space[24][25][26] (the Laplace transform of a function is defined with, \bar{f}(s)=\int_0^{\infty}e^{-st}f(t)dt). Here, the system is defined through the jumping time (JT) PDFs: \psi_{ij}(t) connecting state i with state j (the jump is from state i). The solution is based on the path representation of the Green's function, calculated when including all the path probability density functions of all lengths:


\bar{G}_{ij}(s)= \bar{\Gamma}_{ij}(s)\frac{\bar{\Phi}(s,\tilde{L})}{\bar{\Phi}(s,L)}\bar{\Psi}_i(s).

 

 

 

 

(1)

Here, \bar{\Psi}_{i}(s) =\Sigma_{j}\bar{\Psi}_{ij}(s) and \bar{\Psi}_{ij}(s)=\frac{1-\bar{\psi}_{ij}(s)}{s}. Also, in Eq. (1),


\bar{\Gamma}_{ij}(s)=\Pi_{c=0}^{i\mp1}\bar{\psi}_{c\pm 1c}(s),

 

 

 

 

(2)

and,


\bar{\Phi}(s,L)=1%2B\Sigma_{c=1}^{[L/2]}(-1)^c\bar{h}(s,c;L)

 

 

 

 

(3)

with,


\bar{h}(s,i;L)=\Pi_{c=1}^i\Sigma_{k_c=2%2Bk_{c-1}}^{L-1-2(i-c)}\bar{f}_{k_c}(s)

 

 

 

 

(4)

and,


\bar{f}_{k_j}(s)=\bar{\psi}_{k_jk_j%2B1}(s)\bar{\psi}_{k_j%2B1k_j}(s).

 

 

 

 

(5)

For L=1, \bar{\Phi}(s;L)=1. The symbol [L/2] that appears in the upper bound in the \Sigma in eq. (3) is the floor operation (round towards zero). Finally, the factor \Phi(s,\tilde{L}) in eq. (1) has the same form as in \bar{\Phi}(s;L) in eqs. (3)-(5) , yet it is calculated on a lattice \tilde{L} . Lattice \tilde{L} is constructed from the original lattice by taking out from it the states i and j and the states between them, and then connecting the obtained two fragments. For cases in which a fragment is a single state, this fragment is excluded; namely, lattice \tilde{L} is the longer fragment. When each fragment is a single state, \bar{\Phi}(s;\tilde{L})=1.

Equations (1)-(5) hold for any 1D semi-Markovian random walk in a L-state chain, and form the most general solution in an explicit form for random walks in 1d.

See also

References

  1. ^ Pearson, K. (1905). The problem of the Random Walk. Nature. 72, 294.
  2. ^ Van Kampen N. G., Stochastic Processes in Physics and Chemistry, revised and enlarged edition (North-Holland, Amsterdam) 1992.
  3. ^ Redner S., A Guide to First-Passage Process (Cambridge University Press, Cambridge, UK) 2001.
  4. ^ Goel N. W. and Richter-Dyn N., Stochastic Models in Biology (Academic Press, New York) 1974.
  5. ^ Doi M. and Edwards S. F., The Theory of Polymer Dynamics (Clarendon Press, Oxford) 1986
  6. ^ De Gennes P. G., Scaling Concepts in Polymer Physics (Cornell University Press, Ithaca and London) 1979.
  7. ^ Risken H., The Fokker–Planck Equation (Springer, Berlin) 1984.
  8. ^ Weiss G. H., Aspects and Applications of the Random Walk (North-Holland, Amsterdam) 1994.
  9. ^ Cox D. R., Renewal Theory (Methuen, London) 1962.
  10. ^ Hence the drunkard's random walk would eventually cover all of the city streets (2 Euclidean dimensions) and he will eventually return home, whereas the bird taking a 'random walk' flight through the air (3 Euclidean dimensions) will not cover all space, and will not return to their starting point.
  11. ^ Dana Mackenzie, Taking the Measure of the Wildest Dance on Earth, Science, Vol. 290, no. 5498, pp. 1883–1884.
  12. ^ D. Ben-Avraham and S. Havlin, Diffusion and Reactions in Fractals and Disordered Systems, Cambridge University Press, 2000.
  13. ^ Leo Grady (2006): "Random Walks for Image Segmentation", IEEE Transactions on Pattern Analysis and Machine Intelligence, pp. 1768–1783, Vol. 28, No. 11
  14. ^ Ralf Engbert, Konstantin Mergenthaler, Petra Sinn, and Arkady Pikovsk: [http://www.pnas.org/content/early/2011/08/17/1102730108.full.pdf "An integrated model of fixational eye movements and microsaccades"]
  15. ^ Neal Madras and Gordon Slade (1996), The Self-Avoiding Walk, Birkhäuser Boston. ISBN 0-8176-3891-1.
  16. ^ Gregory Lawler (1996). Intersection of random walks, Birkhäuser Boston. ISBN 0-8176-3892-X.
  17. ^ Gregory Lawler, Conformally Invariant Processes in the Plane, book.ps.
  18. ^ Robin Pemantle (2007), A survey of random processes with reinforcement.
  19. ^ Alamgir, M and von Luxburg, U (2010). "Multi-agent random walks for local clustering on graphs", IEEE 10th International Conference on Data Mining (ICDM), 2010, pp. 18-27.
  20. ^ C.-K. Peng, J. Mietus, J. M. Hausdorff, S. Havlin, H. E. Stanley, A. L. Goldberger (1993). "Long-range anticorrelations and non-gaussian behavior of the heartbeat". Phys. Rev. Lett. 70 (9): 1343–6. doi:10.1103/PhysRevLett.70.1343. PMID 10054352. http://havlin.biu.ac.il/Publications.php?keyword=Long-range+anticorrelations+and+non-gaussian+behavior+of+the+heartbeat&year=*&match=all. 
  21. ^ C.-K. Peng, S. V. Buldyrev, A. L. Goldberger, S. Havlin, F. Sciortino, M. Simons, H. E. Stanley (1992). "Long-range correlations in nucleotide sequences". Nature 365 (6365): 168. doi:10.1038/356168a0. http://havlin.biu.ac.il/Publications.php?keyword=Long-range+correlations+in+nucleotide+sequences&year=*&match=all. 
  22. ^ Y. Liu, P. Cizeau, M. Meyer, C.-K. Peng, H. E. Stanley (1997). "Correlations in economic time series". Physica A 245 (3–4): 437. doi:10.1016/S0378-4371(97)00368-3. 
  23. ^ E. Koscielny-Bunde, A. Bunde, S. Havlin, H. E. Roman, Y. Goldreich, H.-J. Schellenhuber (1998). "Indication of a universal persistence law governing atmospheric variability". Phys. Rev. Lett. 81 (3): 729. doi:10.1103/PhysRevLett.81.729. http://havlin.biu.ac.il/Publications.php?keyword=Indication+of+a+universal+persistence+law+governing+atmospheric+variability&year=*&match=all. 
  24. ^ Flomenbom O. and Klafter J., Phys. Rev. Lett., 95 (2005) 098105-1
  25. ^ Flomenbom O. and Silbey R. J., J. Chem. Phys. 127, 034102 (2007).
  26. ^ Flomenbom O and Silbey RJ, Phys. Rev. E 76, 041101 (2007).

Bibliography

Chapter 3 of this book contains a thorough discussion of random walks, including advanced results, using only elementary tools.

External links